GradPU: Positive-Unlabeled Learning via Gradient Penalty and Positive Upweighting

نویسندگان

چکیده

Positive-unlabeled learning is an essential problem in many real-world applications with only labeled positive and unlabeled data, especially when the negative samples are difficult to identify. Most existing positive-unlabeled methods will inevitably overfit class some extent due existence of unidentified samples. This paper first analyzes overfitting proposes bound generalization errors via Wasserstein distances. Based on that, we develop a simple yet effective method, GradPU, which consists two key ingredients: A gradient-based regularizer that penalizes gradient norms interpolated data region, improves class; An unnormalized upweighting mechanism assigns larger weights those hard, not-well-fitted less frequently labeled. It enforces training error each sample be small increases robustness labeling bias. We evaluate our proposed GradPU three datasets: MNIST, FashionMNIST, CIFAR10. The results demonstrate achieves state-of-the-art performance both unbiased biased scenarios.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Multi-Positive and Unlabeled Learning

Yixing Xu†, Chang Xu‡, Chao Xu†, Dacheng Tao‡ †Key Laboratory of Machine Perception (MOE), Cooperative Medianet Innovation Center, School of Electronics Engineering and Computer Science, PKU, Beijing 100871, China ‡UBTech Sydney AI Institute, The School of Information Technologies, The University of Sydney, J12, 1 Cleveland St, Darlington, NSW 2008, Australia [email protected], [email protected]...

متن کامل

Theoretical Comparisons of Positive-Unlabeled Learning against Positive-Negative Learning

In PU learning, a binary classifier is trained from positive (P) and unlabeled (U) data without negative (N) data. Although N data is missing, it sometimes outperforms PN learning (i.e., ordinary supervised learning). Hitherto, neither theoretical nor experimental analysis has been given to explain this phenomenon. In this paper, we theoretically compare PU (and NU) learning against PN learning...

متن کامل

Positive unlabeled learning via wrapper-based adaptive sampling

Learning from positive and unlabeled data frequently occurs in applications where only a subset of positive instances is available while the rest of the data are unlabeled. In such scenarios, often the goal is to create a discriminant model that can accurately classify both positive and negative data by modelling from labeled and unlabeled instances. In this study, we propose an adaptive sampli...

متن کامل

Positive and unlabeled learning in categorical data

In common binary classification scenarios, the presence of both positive and negative examples in training data is needed to build an efficient classifier. Unfortunately, in many domains, this requirement is not satisfied and only one class of examples is available. To cope with this setting, classification algorithms have been introduced that learn from Positive and Unlabeled (PU) data. Origin...

متن کامل

Positive and Unlabeled Examples Help Learning

In many learning problems, labeled examples are rare or expensive while numerous unlabeled and positive examples are available. However, most learning algorithms only use labeled examples. Thus we address the problem of learning with the help of positive and unlabeled data given a small number of labeled examples. We present both theoretical and empirical arguments showing that learning algorit...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2023

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v37i6.25889